Learning for Video Compression With Recurrent Auto-Encoder and Recurrent Probability Model
نویسندگان
چکیده
The past few years have witnessed increasing interests in applying deep learning to video compression. However, the existing approaches compress a frame with only number of reference frames, which limits their ability fully exploit temporal correlation among frames. To overcome this shortcoming, paper proposes Recurrent Learned Video Compression (RLVC) approach Auto-Encoder (RAE) and Probability Model (RPM). Specifically, RAE employs recurrent cells both encoder decoder. As such, information large range frames can be used for generating latent representations reconstructing compressed outputs. Furthermore, proposed RPM network recurrently estimates Mass Function (PMF) representation, conditioned on distribution previous representations. Due consecutive conditional cross entropy lower than independent entropy, thus reducing bit-rate. experiments show that our achieves state-of-the-art learned compression performance terms PSNR MS-SSIM. Moreover, outperforms default Low-Delay P (LDP) setting x265 PSNR, also has better MS-SSIM SSIM-tuned slowest x265. codes are available at https://github.com/RenYang-home/RLVC.git.
منابع مشابه
Texture Synthesis with Recurrent Variational Auto-Encoder
We propose a recurrent variational auto-encoder for texture synthesis. A novel loss function, FLTBNK, is used for training the texture synthesizer. It is rotational and partially color invariant loss function. Unlike L2 loss, FLTBNK explicitly models the correlation of color intensity between pixels. Our texture synthesizer 1 generates neighboring tiles to expand a sample texture and is evaluat...
متن کاملAuto-encoder pre-training of segmented-memory recurrent neural networks
The extended Backpropagation Through Time (eBPTT) learning algorithm for Segmented-Memory Recurrent Neural Networks (SMRNNs) yet lacks the ability to reliably learn long-term dependencies. The alternative learning algorithm, extended Real-Time Recurrent Learning (eRTRL), does not suffer this problem but is computational very intensive, such that it is impractical for the training of large netwo...
متن کاملVariational Recurrent Auto-Encoders
In this paper we propose a model that combines the strengths of RNNs and SGVB: the Variational Recurrent Auto-Encoder (VRAE). Such a model can be used for efficient, large scale unsupervised learning on time series data, mapping the time series data to a latent vector representation. The model is generative, such that data can be generated from samples of the latent space. An important contribu...
متن کاملHierarchical Recurrent Neural Encoder for Video Representation with Application to Captioning – Supplementary Material
To choose proper parameters for our method, we perform experiments with different parameter settings. We first study how to choose the number of hidden units and report the results in Table 1. For simplification, we set the number of hidden units in LSTMs from the encoder and the decoder as the same. The result shows that the performance of our method can be improved by increasing the number of...
متن کاملLearning Temporal Evolution of Probability Distribution with Recurrent Neural Network
We propose to tackle a time series regression problem by computing temporal evolution of a probability density function to provide a probabilistic forecast. A Recurrent Neural Network (RNN) based model is employed to learn a nonlinear operator for temporal evolution of a probability density function. We use a softmax layer for a numerical discretization of a smooth probability density functions...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Journal of Selected Topics in Signal Processing
سال: 2021
ISSN: ['1941-0484', '1932-4553']
DOI: https://doi.org/10.1109/jstsp.2020.3043590